📚 node [[downsampling|downsampling]]
Welcome! Nobody has contributed anything to 'downsampling|downsampling' yet. You can:
-
Write something in the document below!
- There is at least one public document in every node in the Agora. Whatever you write in it will be integrated and made available for the next visitor to read and edit.
- Write to the Agora from social media.
-
Sign up as a full Agora user.
- As a full user you will be able to contribute your personal notes and resources directly to this knowledge commons. Some setup required :)
⥅ related node [[downsampling]]
⥅ node [[downsampling]] pulled by Agora
📓
garden/KGBicheno/Artificial Intelligence/Introduction to AI/Week 3 - Introduction/Definitions/Downsampling.md by @KGBicheno
downsampling
Go back to the [[AI Glossary]]
Overloaded term that can mean either of the following:
Reducing the amount of information in a feature in order to train a model more efficiently. For example, before training an image recognition model, downsampling high-resolution images to a lower-resolution format.
Training on a disproportionately low percentage of over-represented class examples in order to improve model training on under-represented classes. For example, in a class-imbalanced dataset, models tend to learn a lot about the majority class and not enough about the minority class. Downsampling helps balance the amount of training on the majority and minority classes.
📖 stoas
- public document at doc.anagora.org/downsampling|downsampling
- video call at meet.jit.si/downsampling|downsampling
🔎 full text search for 'downsampling|downsampling'